Product Code Database
Example Keywords: wii -mario $27-188
barcode-scavenger
   » » Wiki: Neural Processing Unit
Tag Wiki 'Neural Processing Unit'.
Tag

A neural processing unit ( NPU), also known as AI accelerator or deep learning processor, is a class of specialized hardware accelerator or computer system designed to accelerate artificial intelligence (AI) and applications, including artificial neural networks and .


Use
Their purpose is either to efficiently execute already trained AI models (inference) or to train AI models. Their applications include for , Internet of things, and data-intensive or sensor-driven tasks. Google using its own AI accelerators. They are often manycore or spatial designs and focus on low-precision arithmetic, novel dataflow architectures, or in-memory computing capability. , a typical datacenter-grade AI integrated circuit chip, the H100 GPU, of .


Consumer devices
AI accelerators are used in mobile devices such as Apple , AMD
(2023). 9781450394178, Association for Computing Machinery.
in Versal and NPUs, , and smartphones, and seen in many , , , and smartphone processors.

It is more recently (circa 2022) added to computer processors from , , and Apple silicon. All models of Intel processors have a built-in versatile processor unit ( VPU) for accelerating inference for computer vision and deep learning.

On consumer devices, the NPU is intended to be small, power-efficient, but reasonably fast when used to run small models. To do this they are designed to support low-bitwidth operations using data types such as INT4, INT8, FP8, and FP16. A common metric is trillions of operations per second (TOPS), though this metric alone does not quantify which kind of operations are being done.


Datacenters
Accelerators are used in servers, including tensor processing units (TPU) in Google Cloud Platform and and chips in Amazon Web Services. Many vendor-specific terms exist for devices in this category, and it is an emerging technology without a .

Since the late 2010s, graphics processing units designed by companies such as and often include AI-specific hardware in the form of dedicated functional units for low-precision matrix-multiplication operations. These GPUs are commonly used as AI accelerators, both for and .


Programming
Mobile NPU vendors typically provide their own application programming interface such as the Snapdragon Neural Processing Engine. An operating system or a higher-level library may provide a more generic interface such as TensorFlow Lite with LiteRT Next (Android) or CoreML (iOS, macOS).

Consumer CPU-integrated NPUs are accessible through vendor-specific APIs. AMD (Ryzen AI), Intel (OpenVINO), Apple silicon (CoreML) each have their own APIs, which can be built upon by a higher-level library.

GPUs generally use existing pipelines such as and adapted for lower precisions. Custom-built systems such as the Google TPU use private interfaces.


Notes

External links

Page 1 of 1
1
Page 1 of 1
1

Account

Social:
Pages:  ..   .. 
Items:  .. 

Navigation

General: Atom Feed Atom Feed  .. 
Help:  ..   .. 
Category:  ..   .. 
Media:  ..   .. 
Posts:  ..   ..   .. 

Statistics

Page:  .. 
Summary:  .. 
1 Tags
10/10 Page Rank
5 Page Refs